# Semantic similarity calculation
Medical Embedded V4
Apache-2.0
This is a multilingual sentence embedding model that can map sentences and paragraphs to a 768-dimensional vector space, suitable for tasks such as clustering and semantic search.
Text Embedding Supports Multiple Languages
M
shtilev
202
1
Langcache Crossencoder V1 Ms Marco MiniLM L12 V2
Apache-2.0
A CrossEncoder model based on the Transformer architecture, fine-tuned on the Quora question pair dataset, used to calculate scores for text pairs, suitable for semantic similarity and semantic search tasks.
Text Classification English
L
aditeyabaral-redis
281
0
Langcache Crossencoder V1 Ms Marco MiniLM L6 V2
Apache-2.0
This is a model based on the Cross Encoder architecture, specifically designed for text pair classification tasks. It is fine-tuned on the Quora question pair dataset and is suitable for semantic similarity judgment and semantic search scenarios.
Text Classification English
L
aditeyabaral-redis
338
0
Langcache Embed V2
A sentence transformer model fine-tuned based on Redis Langcache Embed v1, used to generate 768-dimensional sentence embedding vectors
Text Embedding
L
redis
126
1
Dragonkue KoEn E5 Tiny
Apache-2.0
This is a sentence-transformers model fine-tuned from intfloat/multilingual-e5-small, trained with Korean query-passage pairs to enhance performance in Korean retrieval tasks.
Text Embedding Supports Multiple Languages
D
exp-models
607
5
All MiniLM L2 V2
Apache-2.0
This model is distilled from all-MiniLM-L12-v2, achieving nearly 2x faster inference speed while maintaining high accuracy on both CPU and GPU.
Text Embedding Supports Multiple Languages
A
tabularisai
5,063
2
Snowflake Arctic Embed L V2.0 Ko
Apache-2.0
This is a SentenceTransformer model fine-tuned from Snowflake/snowflake-arctic-embed-l-v2.0, trained on a clustering dataset. It maps sentences and paragraphs into a 1024-dimensional dense vector space, suitable for semantic text similarity and semantic search.
Text Embedding Supports Multiple Languages
S
dragonkue
4,964
26
Jina Embeddings V3
Jina Embeddings V3 is a multilingual sentence embedding model supporting over 100 languages, focusing on sentence similarity calculation and feature extraction tasks.
Text Embedding
Transformers Supports Multiple Languages

J
Daxtra
55
1
Context Skill Extraction Base
This is a model trained based on sentence-transformers, capable of mapping sentences and paragraphs into a 768-dimensional dense vector space, suitable for various tasks such as semantic text similarity calculation and semantic search.
Text Embedding
C
TechWolf
189
5
Gte Base Ko
A sentence embedding model fine-tuned on the Korean triplet dataset based on the Alibaba-NLP/gte-multilingual-base model for semantic similarity calculation
Text Embedding Supports Multiple Languages
G
juyoungml
18
2
Mind Map Blog Model
This is a sentence transformer model fine - tuned from sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2, which can map text to a 384 - dimensional vector space for tasks such as semantic similarity calculation.
Text Embedding
M
hothanhtienqb
463
2
Stella En 400M V5 Cpu
MIT
stella_en_400M_v5_cpu is a model that performs excellently in multiple natural language processing tasks, especially in tasks such as classification, retrieval, clustering, and semantic text similarity.
Text Embedding
S
biggunnyso4
612
1
Gte Base Korean
Apache-2.0
A Korean sentence embedding model fine - tuned on Alibaba - NLP/gte - multilingual - base, supporting tasks such as semantic text similarity calculation and semantic search.
Text Embedding
G
upskyy
1,436
4
E5 All Nli Triplet Matryoshka
This is a sentence-transformers model fine-tuned on intfloat/multilingual-e5-small, designed to map sentences and paragraphs into a 384-dimensional dense vector space, supporting tasks such as semantic text similarity and semantic search.
Text Embedding
E
Omartificial-Intelligence-Space
14
2
USER Bge M3
Apache-2.0
Russian universal sentence encoder, based on the sentence-transformers framework, specifically designed to extract 1024-dimensional dense vectors for Russian text
Text Embedding Other
U
deepvk
339.46k
58
Labse Ru Turbo
MIT
BERT model for computing Russian sentence embeddings, developed based on cointegrated/LaBSE-en-ru with optimized Russian language processing performance
Text Embedding
Transformers Other

L
sergeyzh
3,987
15
Jina Embeddings V2 Base Zh
Apache-2.0
Jina Embeddings V2 Base is a sentence embedding model optimized for Chinese, which can convert text into high-dimensional vector representations for calculating sentence similarity and feature extraction.
Text Embedding Supports Multiple Languages
J
silverjam
63
1
Llm2vec Meta Llama 3 8B Instruct Mntp Supervised
MIT
LLM2Vec is a supervised learning model based on Meta-Llama-3, focusing on natural language processing tasks such as sentence similarity, and supporting various application scenarios such as text embedding, information retrieval, and text classification.
Large Language Model English
L
McGill-NLP
5,530
49
Sentence Transformers Multilingual E5 Small
MIT
multilingual - e5 - small is a model that performs excellently in multilingual text processing tasks, supporting various tasks such as classification, retrieval, clustering, re - ranking, and semantic text similarity.
Text Embedding Supports Multiple Languages
S
beademiguelperez
3,922
1
Ruropebert Classic Base 512
A Russian encoder model based on the RoPEBert architecture, trained using cloning methods, supports 512-token context, and surpasses the original ruBert-base model in quality
Large Language Model
Transformers Other

R
Tochka-AI
103
1
Sambert
This is a Hebrew embedding model based on sentence-transformers, capable of mapping sentences and paragraphs into a 768-dimensional dense vector space, suitable for tasks such as clustering or semantic search.
Text Embedding
Transformers Other

S
MPA
149
2
Multi Sentence BERTino
MIT
This is a sentence transformer model based on BERTino, capable of mapping Italian sentences and paragraphs into a 768-dimensional dense vector space, suitable for tasks such as clustering or semantic search.
Text Embedding
Transformers Other

M
nickprock
63.88k
5
Simcse Roberta Large Zh
MIT
SimCSE(sup) is a model for Chinese sentence similarity tasks. It can encode sentences into embedding vectors and calculate the cosine similarity between sentences.
Text Embedding
Transformers Chinese

S
hellonlp
179
1
Klue Roberta Base Klue Sts
This is a model based on sentence-transformers that can map sentences and paragraphs to a 768-dimensional dense vector space, suitable for tasks such as clustering and semantic search.
Text Embedding
K
shangrilar
165
0
Robbert 2022 Dutch Sentence Transformers Onnx
ONNX version of the Dutch sentence transformer based on the RobBERT model, mapping text to a 768-dimensional vector space, suitable for semantic search and clustering tasks
Text Embedding
Transformers Other

R
Todai
30
1
Sentence Transformers Alephbertgimmel Small
This is a Hebrew sentence similarity calculation model based on sentence-transformers, which can map text to a 512-dimensional vector space for semantic search and clustering tasks
Text Embedding
Transformers Other

S
imvladikon
39
1
Sup Simcse Ja Large
This is a Japanese sentence embedding model trained using the supervised SimCSE method, specifically designed for generating high-quality sentence representations.
Text Embedding
Transformers Japanese

S
cl-nagoya
2,315
14
Bge Base En V1.5 Ct2
MIT
BGE Base English v1.5 is a transformer-based sentence embedding model, specifically designed for extracting sentence features and calculating sentence similarity.
Text Embedding
Transformers English

B
winstxnhdw
30
0
E5 Base En Ru
MIT
This is a vocabulary-pruned version of intfloat/multilingual-e5-base, retaining only English and Russian vocabulary.
Text Embedding
Transformers Supports Multiple Languages

E
d0rj
733
8
Sentence Transformers Gte Large
This is a sentence embedding model based on sentence-transformers, capable of converting text into 1024-dimensional dense vector representations, suitable for tasks like semantic search and text clustering.
Text Embedding
S
embaas
106
1
Distiluse Base Multilingual Cased V2
Apache-2.0
This is a multilingual sentence embedding model that maps text to a 512-dimensional vector space, suitable for semantic search and clustering tasks.
Text Embedding
Transformers Other

D
lorenpe2
32
0
Indosbert Large
indoSBERT-large is an Indonesian sentence embedding model based on sentence-transformers, which maps sentences and paragraphs into a 256-dimensional vector space, suitable for tasks such as clustering and semantic search.
Text Embedding Other
I
denaya
510
13
Compositional Bert Large Uncased
Apache-2.0
CompCSE and SimCSE are contrastive learning-based sentence embedding models for calculating sentence similarity.
Text Embedding
Transformers English

C
perceptiveshawty
754
2
Text2vec Base Chinese Sentence
Apache-2.0
A Chinese sentence embedding model based on the CoSENT (Cosine Sentence) model, mapping sentences to a 768-dimensional dense vector space, suitable for tasks such as sentence embedding, text matching, or semantic search.
Text Embedding
Transformers Chinese

T
shibing624
1,895
54
Xl Lexeme
A model based on sentence-transformers for mapping target words in sentences to a 1024-dimensional vector space, supporting word similarity calculation and semantic search tasks.
Text Embedding
Transformers

X
pierluigic
1,350
1
Abstract Sim Query
A model that maps abstract sentence descriptions to matching sentences, trained on Wikipedia using a dual-encoder architecture.
Text Embedding
Transformers English

A
biu-nlp
53
12
Congen WangchanBERT Small
Apache-2.0
This is a sentence embedding model based on the ConGen framework, capable of mapping sentences to a 128-dimensional dense vector space, suitable for tasks such as semantic search.
Text Embedding
Transformers

C
kornwtp
812
0
Sentence Transformers Paraphrase Multilingual Mpnet Base V2
Apache-2.0
Multilingual sentence embedding model that maps text to a 768-dimensional vector space, suitable for semantic search and clustering tasks
Text Embedding
Transformers

S
tgsc
17
1
Polish Sts V2
This is a Polish-language sentence embedding model capable of mapping sentences and paragraphs into a 1024-dimensional vector space, suitable for semantic search and clustering tasks.
Text Embedding
Transformers Other

P
radlab
43
2
Sentence Transformer Ult5 Pt Small
A sentence transformer model based on ult5-pt-small that maps sentences and paragraphs into 512-dimensional vectors, suitable for tasks like text clustering, similarity calculation, and semantic search.
Text Embedding
Transformers

S
tgsc
358
2
- 1
- 2
- 3
Featured Recommended AI Models